3,294 research outputs found

    Automatic heterogeneous quantization of deep neural networks for low-latency inference on the edge for particle detectors

    Full text link
    Although the quest for more accurate solutions is pushing deep learning research towards larger and more complex algorithms, edge devices demand efficient inference and therefore reduction in model size, latency and energy consumption. One technique to limit model size is quantization, which implies using fewer bits to represent weights and biases. Such an approach usually results in a decline in performance. Here, we introduce a method for designing optimally heterogeneously quantized versions of deep neural network models for minimum-energy, high-accuracy, nanosecond inference and fully automated deployment on chip. With a per-layer, per-parameter type automatic quantization procedure, sampling from a wide range of quantizers, model energy consumption and size are minimized while high accuracy is maintained. This is crucial for the event selection procedure in proton-proton collisions at the CERN Large Hadron Collider, where resources are strictly limited and a latency of O(1) μ{\mathcal O}(1)~\mus is required. Nanosecond inference and a resource consumption reduced by a factor of 50 when implemented on field-programmable gate array hardware are achieved

    Fast convolutional neural networks on FPGAs with hls4ml

    Get PDF
    We introduce an automated tool for deploying ultra low-latency, low-power deep neural networks with convolutional layers on FPGAs. By extending the hls4ml library, we demonstrate an inference latency of 5 μ5\,\mus using convolutional architectures, targeting microsecond latency applications like those at the CERN Large Hadron Collider. Considering benchmark models trained on the Street View House Numbers Dataset, we demonstrate various methods for model compression in order to fit the computational constraints of a typical FPGA device used in trigger and data acquisition systems of particle detectors. In particular, we discuss pruning and quantization-aware training, and demonstrate how resource utilization can be significantly reduced with little to no loss in model accuracy. We show that the FPGA critical resource consumption can be reduced by 97% with zero loss in model accuracy, and by 99% when tolerating a 6% accuracy degradation.Comment: 18 pages, 18 figures, 4 table

    Combination of searches for heavy resonances decaying to WW, WZ, ZZ, WH, and ZH boson pairs in proton–proton collisions at s\sqrt{s} = 8 and 13 TeV

    Full text link
    A statistical combination of searches is presented for massive resonances decaying to WW, WZ, ZZ, WH, and ZH boson pairs in proton–proton collision data collected by the CMS experiment at the LHC. The data were taken at centre-of-mass energies of 8 and 13 TeV, corresponding to respective integrated luminosities of 19.7 and up to 2.7 fb−1fb^{−1}. The results are interpreted in the context of heavy vector triplet and singlet models that mimic properties of composite-Higgs models predicting W′ and Z′ bosons decaying to WZ, WW, WH, and ZH bosons. A model with a bulk graviton that decays into WW and ZZ is also considered. This is the first combined search for WW, WZ, WH, and ZH resonances and yields lower limits on masses at 95% confidence level for W′ and Z′ singlets at 2.3 TeV, and for a triplet at 2.4 TeV. The limits on the production cross section of a narrow bulk graviton resonance with the curvature scale of the warped extra dimension k~=0.5\tilde{k}=0.5, in the mass range of 0.6 to 4.0 TeV, are the most stringent published to date

    Search for supersymmetry in pp collisions at s\sqrt{s} = 13 tev in the single-lepton final state using the sum of masses of large-radius jets

    Full text link
    Results are reported from a search for supersymmetric particles in proton-proton collisions in the final state with a single lepton, multiple jets, including at least one b-tagged jet, and large missing transverse momentum. The search uses a sample of proton-proton collision data at s\sqrt{s} = 13 TeV recorded by the CMS experiment at the LHC, corresponding to an integrated luminosity of 35.9  fb−1fb^{−1}. The observed event yields in the signal regions are consistent with those expected from standard model backgrounds. The results are interpreted in the context of simplified models of supersymmetry involving gluino pair production, with gluino decay into either on- or off-mass-shell top squarks. Assuming that the top squarks decay into a top quark plus a stable, weakly interacting neutralino, scenarios with gluino masses up to about 1.9 TeV are excluded at 95% confidence level for neutralino masses up to about 1 TeV

    Search for dark matter produced in association with heavy-flavor quark pairs in proton-proton collisions at s\sqrt{s} = 13 TeV

    Full text link
    A search is presented for an excess of events with heavy-flavor quark pairs (tt‾t\overline{t} and bb‾b\overline{b}) and a large imbalance in transverse momentum in data from proton–proton collisions at a center-of-mass energy of 13TeV. The data correspond to an integrated luminosity of 2.2 fb−1fb^{−1} collected with the CMS detector at the CERN LHC. No deviations are observed with respect to standard model predictions. The results are used in the first interpretation of dark matter production in tt‾t\overline{t} and bb‾b\overline{b} final states in a simplified model. This analysis is also the first to perform a statistical combination of searches for dark matter produced with different heavy-flavor final states. The combination provides exclusions that are stronger than those achieved with individual heavy-flavor final states

    Accelerated Charged Particle Tracking with Graph Neural Networks on FPGAs

    Full text link
    We develop and study FPGA implementations of algorithms for charged particle tracking based on graph neural networks. The two complementary FPGA designs are based on OpenCL, a framework for writing programs that execute across heterogeneous platforms, and hls4ml, a high-level-synthesis-based compiler for neural network to firmware conversion. We evaluate and compare the resource usage, latency, and tracking performance of our implementations based on a benchmark dataset. We find a considerable speedup over CPU-based execution is possible, potentially enabling such algorithms to be used effectively in future computing workflows and the FPGA-based Level-1 trigger at the CERN Large Hadron Collider.Comment: 8 pages, 4 figures, To appear in Third Workshop on Machine Learning and the Physical Sciences (NeurIPS 2020

    The CMS trigger system

    Full text link
    This paper describes the CMS trigger system and its performance during Run 1 of the LHC. The trigger system consists of two levels designed to select events of potential physics interest from a GHz (MHz) interaction rate of proton-proton (heavy ion) collisions. The first level of the trigger is implemented in hardware, and selects events containing detector signals consistent with an electron, photon, muon, Ï„ lepton, jet, or missing transverse energy. A programmable menu of up to 128 object-based algorithms is used to select events for subsequent processing. The trigger thresholds are adjusted to the LHC instantaneous luminosity during data taking in order to restrict the output rate to 100 kHz, the upper limit imposed by the CMS readout electronics. The second level, implemented in software, further refines the purity of the output stream, selecting an average rate of 400 Hz for offline event storage. The objectives, strategy and performance of the trigger system during the LHC Run 1 are described

    Search for dark matter produced with an energetic jet or a hadronically decaying W or Z boson at s=13\sqrt{s}=13 TeV

    Full text link
    A search for dark matter particles is performed using events with large missing transverse momentum, at least one energetic jet, and no leptons, in proton-proton collisions at s=13\sqrt{s}=13 TeV collected with the CMS detector at the LHC. The data sample corresponds to an integrated luminosity of 12.9 fb−1fb^{−1}. The search includes events with jets from the hadronic decays of a W or Z boson. The data are found to be in agreement with the predicted background contributions from standard model processes. The results are presented in terms of simplified models in which dark matter particles are produced through interactions involving a vector, axial-vector, scalar, or pseudoscalar mediator. Vector and axial-vector mediator particles with masses up to 1.95 TeV, and scalar and pseudoscalar mediator particles with masses up to 100 and 430 GeV respectively, are excluded at 95% confidence level. The results are also interpreted in terms of the invisible decays of the Higgs boson, yielding an observed (expected) 95% confidence level upper limit of 0.44 (0.56) on the corresponding branching fraction. The results of this search provide the strongest constraints on the dark matter pair production cross section through vector and axial-vector mediators at a particle collider. When compared to the direct detection experiments, the limits obtained from this search provide stronger constraints for dark matter masses less than 5, 9, and 550 GeV, assuming vector, scalar, and axial-vector mediators, respectively. The search yields stronger constraints for dark matter masses less than 200 GeV, assuming a pseudoscalar mediator, when compared to the indirect detection results from Fermi-LAT

    Search for a light pseudoscalar Higgs boson produced in association with bottom quarks in pp collisions at s\sqrt{s} = 8 TeV

    Full text link
    A search for a light pseudoscalar Higgs boson (A) produced in association with bottom quarks and decaying into a muon pair is reported. The search uses 19.7 fb−1fb{−1} of proton-proton collisions at a center-of-mass energy of 8 TeV, collected by the CMS experiment. No signal is observed in the dimuon mass range from 25 to 60 GeV. Upper limits on the cross section times branching fraction, σ(pp→bb‾A)B(A→μμ)\sigma (pp \to b\overline{b}A)\mathcal{B}(A \to \mu \mu), are set
    • …
    corecore